lateral inhibition
- North America > United States > Florida > Hillsborough County > Tampa (0.14)
- North America > United States > Pennsylvania (0.04)
- Africa > Mali (0.04)
- (2 more...)
- Education > Educational Setting (0.68)
- Health & Medicine > Therapeutic Area (0.46)
- North America > United States > Florida > Hillsborough County > Tampa (0.14)
- Africa > Mali (0.04)
- North America > United States > Pennsylvania (0.04)
- (4 more...)
- North America > United States > Florida > Hillsborough County > Tampa (0.14)
- North America > United States > Pennsylvania (0.04)
- Africa > Mali (0.04)
- (2 more...)
- Education > Educational Setting (0.68)
- Health & Medicine > Therapeutic Area (0.46)
- North America > United States > Florida > Hillsborough County > Tampa (0.14)
- Africa > Mali (0.04)
- North America > United States > Pennsylvania (0.04)
- (4 more...)
Differential Gated Self-Attention
Lygizou, Elpiniki Maria, Farsang, Mónika, Grosu, Radu
Transformers excel across a large variety of tasks but remain susceptible to corrupted inputs, since standard self-attention treats all query-key interactions uniformly. Inspired by lateral inhibition in biological neural circuits and building on the recent use by the Differential Transformer's use of two parallel softmax subtraction for noise cancellation, we propose Multihead Differential Gated Self-Attention (M-DGSA) that learns per-head input-dependent gating to dynamically suppress attention noise. Each head splits into excitatory and inhibitory branches whose dual softmax maps are fused by a sigmoid gate predicted from the token embedding, yielding a context-aware contrast enhancement. M-DGSA integrates seamlessly into existing Transformer stacks with minimal computational overhead. We evaluate on both vision and language benchmarks, demonstrating consistent robustness gains over vanilla Transformer, Vision Transformer, and Differential Transformer baselines. Our contributions are (i) a novel input-dependent gating mechanism for self-attention grounded in lateral inhibition, (ii) a principled synthesis of biological contrast-enhancement and self-attention theory, and (iii) comprehensive experiments demonstrating noise resilience and cross-domain applicability.
- Europe > Austria > Vienna (0.14)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- North America > United States > Oregon > Multnomah County > Portland (0.04)
- North America > Canada > Ontario > Toronto (0.04)
- Information Technology (0.46)
- Leisure & Entertainment (0.46)
- Health & Medicine (0.46)
Growing Artificial Neural Networks for Control: the Role of Neuronal Diversity
Nisioti, Eleni, Plantec, Erwan, Montero, Milton, Pedersen, Joachim Winther, Risi, Sebastian
In biological evolution complex neural structures grow from a handful of cellular ingredients. As genomes in nature are bounded in size, this complexity is achieved by a growth process where cells communicate locally to decide whether to differentiate, proliferate and connect with other cells. This self-organisation is hypothesized to play an important part in the generalisation, and robustness of biological neural networks. Artificial neural networks (ANNs), on the other hand, are traditionally optimized in the space of weights. Thus, the benefits and challenges of growing artificial neural networks remain understudied. Building on the previously introduced Neural Developmental Programs (NDP), in this work we present an algorithm for growing ANNs that solve reinforcement learning tasks. We identify a key challenge: ensuring phenotypic complexity requires maintaining neuronal diversity, but this diversity comes at the cost of optimization stability. To address this, we introduce two mechanisms: (a) equipping neurons with an intrinsic state inherited upon neurogenesis; (b) lateral inhibition, a mechanism inspired by biological growth, which controlls the pace of growth, helping diversity persist. We show that both mechanisms contribute to neuronal diversity and that, equipped with them, NDPs achieve comparable results to existing direct and developmental encodings in complex locomotion tasks
- North America > United States (0.14)
- Oceania > Australia > Victoria > Melbourne (0.05)
- Europe > Denmark > Capital Region > Copenhagen (0.04)
- (2 more...)
End-to-End Lip Reading in Romanian with Cross-Lingual Domain Adaptation and Lateral Inhibition
Mănescu, Emilian-Claudiu, Smădu, Răzvan-Alexandru, Avram, Andrei-Marius, Cercel, Dumitru-Clementin, Pop, Florin
Lip reading or visual speech recognition has gained significant attention in recent years, particularly because of hardware development and innovations in computer vision. While considerable progress has been obtained, most models have only been tested on a few large-scale datasets. This work addresses this shortcoming by analyzing several architectures and optimizations on the underrepresented, short-scale Romanian language dataset called Wild LRRo. Most notably, we compare different backend modules, demonstrating the effectiveness of adding ample regularization methods. We obtain state-of-the-art results using our proposed method, namely cross-lingual domain adaptation and unlabeled videos from English and German datasets to help the model learn language-invariant features. Lastly, we assess the performance of adding a layer inspired by the neural inhibition mechanism.
- Europe > Romania > București - Ilfov Development Region > Municipality of Bucharest > Bucharest (0.05)
- Asia > Middle East > Qatar > Ad-Dawhah > Doha (0.04)
- North America > United States > Washington > King County > Seattle (0.04)
- (6 more...)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Cognitive Science (1.00)
- Information Technology > Artificial Intelligence > Vision (0.89)
Multilingual Multiword Expression Identification Using Lateral Inhibition and Domain Adaptation
Avram, Andrei-Marius, Mititelu, Verginica Barbu, Păiş, Vasile, Cercel, Dumitru-Clementin, Trăuşan-Matu, Ştefan
Correctly identifying multiword expressions (MWEs) is an important task for most natural language processing systems since their misidentification can result in ambiguity and misunderstanding of the underlying text. In this work, we evaluate the performance of the mBERT model for MWE identification in a multilingual context by training it on all 14 languages available in version 1.2 of the PARSEME corpus. We also incorporate lateral inhibition and language adversarial training into our methodology to create language-independent embeddings and improve its capabilities in identifying multiword expressions. The evaluation of our models shows that the approach employed in this work achieves better results compared to the best system of the PARSEME 1.2 competition, MTLB-STRUCT, on 11 out of 14 languages for global MWE identification and on 12 out of 14 languages for unseen MWE identification. Additionally, averaged across all languages, our best approach outperforms the MTLB-STRUCT system by 1.23% on global MWE identification and by 4.73% on unseen global MWE identification.
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.67)
- Information Technology > Artificial Intelligence > Cognitive Science > Neuroscience (0.63)
Romanian Multiword Expression Detection Using Multilingual Adversarial Training and Lateral Inhibition
Avram, Andrei-Marius, Mititelu, Verginica Barbu, Cercel, Dumitru-Clementin
Multiword expressions are a key ingredient for developing large-scale and linguistically sound natural language processing technology. This paper describes our improvements in automatically identifying Romanian multiword expressions on the corpus released for the PARSEME v1.2 shared task. Our approach assumes a multilingual perspective based on the recently introduced lateral inhibition layer and adversarial training to boost the performance of the employed multilingual language models. With the help of these two methods, we improve the F1-score of XLM-RoBERTa by approximately 2.7% on unseen multiword expressions, the main task of the PARSEME 1.2 edition. In addition, our results can be considered SOTA performance, as they outperform the previous results on Romanian obtained by the participants in this competition.
- South America > Colombia > Meta Department > Villavicencio (0.05)
- Europe > Romania > București - Ilfov Development Region > Municipality of Bucharest > Bucharest (0.04)
- Europe > Middle East > Malta > Northern Region > Western District > Attard (0.04)
- (4 more...)
- Information Technology > Artificial Intelligence > Natural Language > Text Processing (0.69)
- Information Technology > Artificial Intelligence > Cognitive Science > Neuroscience (0.63)
- Information Technology > Artificial Intelligence > Natural Language > Information Retrieval (0.50)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.48)
Learning Winner-take-all Competition Between Groups of Neurons in Lateral Inhibitory Networks
It has long been known that lateral inhibition in neural networks can lead to a winner-take-all competition, so that only a single neuron is active at a steady state. Here we show how to organize lateral inhibition so that groups of neurons compete to be active. Given a collection of poten(cid:173) tially overlapping groups, the inhibitory connectivity is set by a formula that can be interpreted as arising from a simple learning rule. Our analy(cid:173) sis demonstrates that such inhibition generally results in winner-take-all competition between the given groups, with the exception of some de(cid:173) generate cases. In a broader context, the network serves as a particular illustration of the general distinction between permitted and forbidden sets, which was introduced recently.